首页> 外文OA文献 >Parallel Matrix Multiplication on Memristor-Based Computation-in-Memory Architecture
【2h】

Parallel Matrix Multiplication on Memristor-Based Computation-in-Memory Architecture

机译:基于忆阻器的内存中计算架构的并行矩阵乘法

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。
获取外文期刊封面目录资料

摘要

One of the most important constraints of today’s architectures for data-intensive applications is the limited bandwidth due to the memory-processor communication bottleneck. This significantly impacts performance and energy. For instance, the energy consumption share of communication and memoryaccess may exceed 80%. Recently, the concept of Computation-in-Memory (CIM) was proposed, which is based on the integration of storage and computation in the same physical location using a crossbar topology and non-volatile resistive-switching memristor technology. To illustrate the tremendous potential of CIM architecture in exploiting massively parallel computation while reducing the communication overhead, we present a communicationefficient mapping of a large-scale matrix multiplication algorithm on the CIM architecture. The experimental results show that, depending on the matrix size, CIM architecture exhibits several orders of magnitude higher performance in total execution timeand two orders of magnitude better in total energy consumption than the multicore-based on the shared memory architecture.
机译:当今数据密集型应用程序体系结构最重要的限制之一是由于内存处理器通信瓶颈而导致带宽有限。这会显着影响性能和能源。例如,通信和内存访问的能耗份额可能超过80%。最近,提出了内存计算(CIM)的概念,该概念基于使用交叉开关拓扑和非易失性电阻开关忆阻器技术在同一物理位置存储和计算的集成。为了说明CIM体系结构在利用大规模并行计算的同时减少通信开销的巨大潜力,我们提出了CIM体系结构上大规模矩阵乘法算法的通信效率映射。实验结果表明,与基于共享内存架构的多核相比,根据矩阵大小,CIM架构的总执行时间性能要高几个数量级,总能耗要高出两个数量级。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号